1,980 research outputs found

    Development of an ontology supporting failure analysis of surface safety valves used in Oil & Gas applications

    Get PDF
    Treball desenvolupat dins el marc del programa 'European Project Semester'.The project describes how to apply Root Cause Analysis (RCA) in the form of a Failure Mode Effect and Criticality Analysis (FMECA) on hydraulically actuated Surface Safety Valves (SSVs) of Xmas trees in oil and gas applications, in order to be able to predict the occurrence of failures and implement preventive measures such as Condition and Performance Monitoring (CPM) to improve the life-span of a valve and decrease maintenance downtime. In the oil and gas industry, valves account for 52% of failures in the system. If these failures happen unexpectedly it can cause a lot of problems. Downtime of the oil well quickly becomes an expensive problem, unscheduled maintenance takes a lot of extra time and the lead-time for replacement parts can be up to 6 months. This is why being able to predict these failures beforehand is something that can bring a lot of benefits to a company. To determine the best course of action to take in order to be able to predict failures, a FMECA report is created. This is an analysis where all possible failures of all components are catalogued and given a Risk Priority Number (RPN), which has three variables: severity, detectability and occurrence. Each of these is given a rating between 0 and 10 and then the variables are multiplied with each other, resulting in the RPN. The components with an RPN above an acceptable risk level are then further investigated to see how to be able to detect them beforehand and how to mitigate the risk that they pose. Applying FMECA to the SSV mean breaking the system down into its components and determining the function, dependency and possible failures. To this end, the SSV is broken up into three sub-systems: the valve, the actuator and the hydraulic system. The hydraulic system is the sub-system of the SSV responsible for containing, transporting and pressurizing of the hydraulic fluid and in turn, the actuator. It also contains all the safety features, such as pressure pilots, and a trip system in case a problem is detected in the oil line. The actuator is, as the name implies, the sub-system which opens and closes the valve. It is made up of a number of parts such as a cylinder, a piston and a spring. These parts are interconnected in a number of ways to allow the actuator to successfully perform its function. The valve is the actual part of the system which interacts with the oil line by opening and closing. Like the actuator, this sub-system is broken down into a number of parts which work together to perform its function. After breaking down and defining each subsystem on a functional level, a model was created using a functional block diagram. Each component also allows for the defining of dependencies and interactions between the different components and a failure diagram for each component. This model integrates the three sub-systems back into one, creating a complete picture of the entire system which can then be used to determine the effects of different failures in components to the rest of the system. With this model completed we created a comprehensive FMECA report and test the different possible CPM solutions to mitigate the largest risks

    Localization in orchards using Extended Kalman Filter for sensor-fusion - A FroboMind component

    Get PDF
    Making an automated vehicle navigate in rows of orchards is a feature, relevant for automating the plant nursing and cultivation of the trees. To be able to navigate accurate and reliably, the vehicle must know its position relative to the trees in the orchards

    Succinct Data Structures for Retrieval and Approximate Membership

    Get PDF
    The retrieval problem is the problem of associating data with keys in a set. Formally, the data structure must store a function f: U ->{0,1}^r that has specified values on the elements of a given set S, a subset of U, |S|=n, but may have any value on elements outside S. Minimal perfect hashing makes it possible to avoid storing the set S, but this induces a space overhead of Theta(n) bits in addition to the nr bits needed for function values. In this paper we show how to eliminate this overhead. Moreover, we show that for any k query time O(k) can be achieved using space that is within a factor 1+e^{-k} of optimal, asymptotically for large n. If we allow logarithmic evaluation time, the additive overhead can be reduced to O(log log n) bits whp. The time to construct the data structure is O(n), expected. A main technical ingredient is to utilize existing tight bounds on the probability of almost square random matrices with rows of low weight to have full row rank. In addition to direct constructions, we point out a close connection between retrieval structures and hash tables where keys are stored in an array and some kind of probing scheme is used. Further, we propose a general reduction that transfers the results on retrieval into analogous results on approximate membership, a problem traditionally addressed using Bloom filters. Again, we show how to eliminate the space overhead present in previously known methods, and get arbitrarily close to the lower bound. The evaluation procedures of our data structures are extremely simple (similar to a Bloom filter). For the results stated above we assume free access to fully random hash functions. However, we show how to justify this assumption using extra space o(n) to simulate full randomness on a RAM

    Tight Thresholds for Cuckoo Hashing via XORSAT

    Full text link
    We settle the question of tight thresholds for offline cuckoo hashing. The problem can be stated as follows: we have n keys to be hashed into m buckets each capable of holding a single key. Each key has k >= 3 (distinct) associated buckets chosen uniformly at random and independently of the choices of other keys. A hash table can be constructed successfully if each key can be placed into one of its buckets. We seek thresholds alpha_k such that, as n goes to infinity, if n/m <= alpha for some alpha < alpha_k then a hash table can be constructed successfully with high probability, and if n/m >= alpha for some alpha > alpha_k a hash table cannot be constructed successfully with high probability. Here we are considering the offline version of the problem, where all keys and hash values are given, so the problem is equivalent to previous models of multiple-choice hashing. We find the thresholds for all values of k > 2 by showing that they are in fact the same as the previously known thresholds for the random k-XORSAT problem. We then extend these results to the setting where keys can have differing number of choices, and provide evidence in the form of an algorithm for a conjecture extending this result to cuckoo hash tables that store multiple keys in a bucket.Comment: Revision 3 contains missing details of proofs, as appendix
    • …
    corecore